A Hierarchical Sparse Matrix Storage Format for Vector Processors
نویسندگان
چکیده
We describe and evaluate a Hierarchical Sparse Matrix (HiSM) storage format designed to be a unified format for sparse matrix applications on vector processors. The advantages that the format offers are low storage requirements, a flexible structure for element manipulations and allowing for efficient operations. To take full advantage of the format we also propose a vector architecture extension that supports the HiSM format. We show that utilizing the HiSM format we can achieve 40% reduction of storage space when comparing to the Compressed Row Storage (CRS) and Jagged Diagonal (JD) storage methods. Utilizing the HiSM storage on a vector processor we can significantly increase the vector performance for Sparse Matrix Vector Multiplication (SMVM) by 5.3 times compared to CRS and 4.07 times compared to JD. Finally, to illustrate the flexibility of the format we compared the cost of an element insertion against the JD and CRS formats. We show that for an element insertion operation HiSM outperforms JD for average and large matrices although it has a slight disadvantage for small matrices and always outperforms CRS between 2 and 400 times.
منابع مشابه
Transposition Mechanism for Sparse Matrices on Vector Processors
Many scientific applications involve operations on sparse matrices. However, due to irregularities induced by the sparsity patterns, many operations on sparse matrices execute inefficiently on traditional scalar and vector architectures. To tackle this problem a scheme has been proposed consisting of two parts: (a) An extension to a vector architecture to support sparse matrix-vector multiplica...
متن کاملData Structures and Algorithms for Distributed Sparse Matrix Operations
We propose extensions of the classical row compressed storage format for sparse matrices. The extensions are designed to accomodate distributed storage of the matrix. We outline an implementation of the matrix-vector product using this distributed storage format, and give algorithms for building and using the communication structure between processors.
متن کاملSparse Matrix Storage Format
Operations on Sparse Matrices are the key computational kernels in many scientific and engineering applications. They are characterized with poor substantiated performance. It is not uncommon for microprocessors to gain only 10-20% of their peak floating-point performance when doing sparse matrix computations even when special vector processors have been added as coprocessor facilities. In this...
متن کاملVector ISA Extension for Sparse Matrix-Vector Multiplication
In this paper we introduce a vector ISA extension to facilitate sparse matrix manipulation on vector processors (VPs). First we introduce a new Block Based Compressed Storage (BBCS) format for sparse matrix representation and a Block-wise Sparse Matrix-Vector Multiplication approach. Additionally, we propose two vector instructions, Multiple Inner Product and Accumulate (MIPA) and LoaD Section ...
متن کاملA Unified Sparse Matrix Data Format for Efficient General Sparse Matrix-Vector Multiplication on Modern Processors with Wide SIMD Units | SIAM Journal on Scientific Computing | Vol. 36, No. 5 | Society for Industrial and Applied Mathematics
Sparse matrix-vector multiplication (spMVM) is the most time-consuming kernel in many numerical algorithms and has been studied extensively on all modern processor and accelerator architectures. However, the optimal sparse matrix data storage format is highly hardware-specific, which could become an obstacle when using heterogeneous systems. Also, it is as yet unclear how the wide single instru...
متن کامل